Goto

Collaborating Authors

 Alpharetta


The Wisdom of the Crowd: High-Fidelity Classification of Cyber-Attacks and Faults in Power Systems Using Ensemble and Machine Learning

Abukhousa, Emad, Afroz, Syed Sohail Feroz Syed, Alsaeed, Fahad, Qwbaiban, Abdulaziz, Zonouz, Saman, Meliopoulos, A. P. Sakis

arXiv.org Artificial Intelligence

This paper presents a high-fidelity evaluation framework for machine learning (ML)-based classification of cyber-attacks and physical faults using electromagnetic transient simulations with digital substation emulation at 4.8 kHz. Twelve ML models, including ensemble algorithms and a multi-layer perceptron (MLP), were trained on labeled time-domain measurements and evaluated in a real-time streaming environment designed for sub-cycle responsiveness. The architecture incorporates a cycle-length smoothing filter and confidence threshold to stabilize decisions. Results show that while several models achieved near-perfect offline accuracies (up to 99.9%), only the MLP sustained robust coverage (98-99%) under streaming, whereas ensembles preserved perfect anomaly precision but abstained frequently (10-49% coverage). These findings demonstrate that offline accuracy alone is an unreliable indicator of field readiness and underscore the need for realistic testing and inference pipelines to ensure dependable classification in inverter-based resources (IBR)-rich networks.


Can human clinical rationales improve the performance and explainability of clinical text classification models?

Metzner, Christoph, Gao, Shang, Herrmannova, Drahomira, Hanson, Heidi A.

arXiv.org Artificial Intelligence

AI-driven clinical text classification is vital for explainable automated retrieval of population-level health information. This work investigates whether human-based clinical rationales can serve as additional supervision to improve both performance and explainability of transformer-based models that automatically encode clinical documents. We analyzed 99,125 human-based clinical rationales that provide plausible explanations for primary cancer site diagnoses, using them as additional training samples alongside 128,649 electronic pathology reports to evaluate transformer-based models for extracting primary cancer sites. We also investigated sufficiency as a way to measure rationale quality for pre-selecting rationales. Our results showed that clinical rationales as additional training data can improve model performance in high-resource scenarios but produce inconsistent behavior when resources are limited. Using sufficiency as an automatic metric to preselect rationales also leads to inconsistent results. Importantly, models trained on rationales were consistently outperformed by models trained on additional reports instead. This suggests that clinical rationales don't consistently improve model performance and are outperformed by simply using more reports. Therefore, if the goal is optimizing accuracy, annotation efforts should focus on labeling more reports rather than creating rationales. However, if explainability is the priority, training models on rationale-supplemented data may help them better identify rationale-like features. We conclude that using clinical rationales as additional training data results in smaller performance improvements and only slightly better explainability (measured as average token-level rationale coverage) compared to training on additional reports.


Efficient Multi-Task Learning via Generalist Recommender

Wang, Luyang, Tang, Cangcheng, Zhang, Chongyang, Ruan, Jun, Huang, Kai, Dai, Jason

arXiv.org Artificial Intelligence

Multi-task learning (MTL) is a common machine learning technique that allows the model to share information across different tasks and improve the accuracy of recommendations for all of them. Many existing MTL implementations suffer from scalability issues as the training and inference performance can degrade with the increasing number of tasks, which can limit production use case scenarios for MTL-based recommender systems. Inspired by the recent advances of large language models, we developed an end-to-end efficient and scalable Generalist Recommender (GRec). GRec takes comprehensive data signals by utilizing NLP heads, parallel Transformers, as well as a wide and deep structure to process multi-modal inputs. These inputs are then combined and fed through a newly proposed task-sentence level routing mechanism to scale the model capabilities on multiple tasks without compromising performance. Offline evaluations and online experiments show that GRec significantly outperforms our previous recommender solutions. GRec has been successfully deployed on one of the largest telecom websites and apps, effectively managing high volumes of online traffic every day.


Tera-SpaceCom: GNN-based Deep Reinforcement Learning for Joint Resource Allocation and Task Offloading in TeraHertz Band Space Networks

Hu, Zhifeng, Han, Chong, Gerstacker, Wolfgang, Akyildiz, Ian F.

arXiv.org Artificial Intelligence

Terahertz (THz) space communications (Tera-SpaceCom) is envisioned as a promising technology to enable various space science and communication applications. Mainly, the realm of Tera-SpaceCom consists of THz sensing for space exploration, data centers in space providing cloud services for space exploration tasks, and a low earth orbit (LEO) mega-constellation relaying these tasks to ground stations (GSs) or data centers via THz links. Moreover, to reduce the computational burden on data centers as well as resource consumption and latency in the relaying process, the LEO mega-constellation provides satellite edge computing (SEC) services to directly compute space exploration tasks without relaying these tasks to data centers. The LEO satellites that receive space exploration tasks offload (i.e., distribute) partial tasks to their neighboring LEO satellites, to further reduce their computational burden. However, efficient joint communication resource allocation and computing task offloading for the Tera-SpaceCom SEC network is an NP-hard mixed-integer nonlinear programming problem (MINLP), due to the discrete nature of space exploration tasks and sub-arrays as well as the continuous nature of transmit power. To tackle this challenge, a graph neural network (GNN)-deep reinforcement learning (DRL)-based joint resource allocation and task offloading (GRANT) algorithm is proposed with the target of long-term resource efficiency (RE). Particularly, GNNs learn relationships among different satellites from their connectivity information. Furthermore, multi-agent and multi-task mechanisms cooperatively train task offloading and resource allocation. Compared with benchmark solutions, GRANT not only achieves the highest RE with relatively low latency, but realizes the fewest trainable parameters and the shortest running time.


Advancing Household Robotics: Deep Interactive Reinforcement Learning for Efficient Training and Enhanced Performance

Soni, Arpita, Alla, Sujatha, Dodda, Suresh, Volikatla, Hemanth

arXiv.org Artificial Intelligence

The market for domestic robots made to perform household chores is growing as these robots relieve people of everyday responsibilities. Domestic robots are generally welcomed for their role in easing human labor, in contrast to industrial robots, which are frequently criticized for displacing human workers. But before these robots can carry out domestic chores, they need to become proficient in several minor activities, such as recognizing their surroundings, making decisions, and picking up on human behaviors. Reinforcement learning, or RL, has emerged as a key robotics technology that enables robots to interact with their environment and learn how to optimize their actions to maximize rewards. However, the goal of Deep Reinforcement Learning is to address more complicated, continuous action-state spaces in real-world settings by combining RL with Neural Networks. The efficacy of DeepRL can be further augmented through interactive feedback, in which a trainer offers real-time guidance to expedite the robot's learning process. Nevertheless, the current methods have drawbacks, namely the transient application of guidance that results in repeated learning under identical conditions. Therefore, we present a novel method to preserve and reuse information and advice via Deep Interactive Reinforcement Learning, which utilizes a persistent rule-based system. This method not only expedites the training process but also lessens the number of repetitions that instructors will have to carry out. This study has the potential to advance the development of household robots and improve their effectiveness and efficiency as learners.


Uncertainty-Aware Calibration of a Hot-Wire Anemometer With Gaussian Process Regression

García-Ruiz, Rubén Antonio, Blanco-Claraco, José Luis, López-Martínez, Javier, Callejón-Ferre, Ángel Jesús

arXiv.org Artificial Intelligence

Expensive ultrasonic anemometers are usually required to measure wind speed accurately. The aim of this work is to overcome the loss of accuracy of a low cost hot-wire anemometer caused by the changes of air temperature, by means of a probabilistic calibration using Gaussian Process Regression. Gaussian Process Regression is a non-parametric, Bayesian, and supervised learning method designed to make predictions of an unknown target variable as a function of one or more known input variables. Our approach is validated against real datasets, obtaining a good performance in inferring the actual wind speed values. By performing, before its real use in the field, a calibration of the hot-wire anemometer taking into account air temperature, permits that the wind speed can be estimated for the typical range of ambient temperatures, including a grounded uncertainty estimation for each speed measure.


MotorDNA, Insurtech, Launches Leveraging Data and AI to Save Lives - InsuranceNewsNet

#artificialintelligence

MotorDNA seeks to close the data gap between the Insurers, Consumers, Auto Lenders, Fleet Managers, Regulators, and OEMs with a platform to encourage the production and demand for safer vehicles. The US Automobile Insurance industry, valued at greater than $300B, is modernizing and transforming its products to mirror the transformation in the mobility space. Automobile technology is rapidly advancing, and vehicles are getting smarter. Safety features, sometimes referred to as ADAS (Advanced Driver Assistance Systems) and their effectiveness are progressing quickly on the journey to autonomous driving. However, vehicle build data is neither readily available in the market nor has it been organized to make it actionable for insurers to create products and pricing based on accurately assessing the risk of each vehicle.


MotorDNA, Insurtech, Launches Leveraging Data and AI to Save Lives

#artificialintelligence

MotorDNA seeks to close the data gap between the Insurers, Consumers, Auto Lenders, Fleet Managers, Regulators, and OEMs with a platform to encourage the production and demand for safer vehicles. The US Automobile Insurance industry, valued at greater than $300B, is modernizing and transforming its products to mirror the transformation in the mobility space. Automobile technology is rapidly advancing, and vehicles are getting smarter. Safety features, sometimes referred to as ADAS (Advanced Driver Assistance Systems) and their effectiveness are progressing quickly on the journey to autonomous driving. However, vehicle build data is neither readily available in the market nor has it been organized to make it actionable for insurers to create products and pricing based on accurately assessing the risk of each vehicle.


Anti-fraud technology with a human touch

#artificialintelligence

The use of artificial intelligence and machine learning in bank fraud analytics is continuing to move from reactively mitigating fraud that's already occurred to preventing fraudulent activities from actually happening--but in ways that try not to block legitimate customer transactions. As anti-fraud technology has become more advanced and scalable, some banks are now investing in a cross-product, omnichannel view of customer behavior, says Philippe Guiral, who leads Accenture's North America fraud and financial crime practice. This means leveraging customer data across domains within the organization to gain more insights of customer behavior to better assess whether any particular transaction is suspicious. A growing number of banks are now building cases to show these solutions can not only improve fraud prevention rates, but also enhance the customer experience and be applied across additional functions--including financial crime, 'Know Your Customer,' risk and customer intelligence--to uncover hidden risks and discover new opportunities, he says. Indeed, it's critical to have a strong fraud analytics solution that can give banks a comprehensive view of a customer's identity and real-time insights into application activity, says Kimberly White, senior director of fraud & identity at LexisNexis Risk Solutions in Alpharetta, Georgia.


Artificial Intelligence: Advancing Applications in the CPI - Chemical Engineering

#artificialintelligence

As data accessibility and analysis capabilities have rapidly advanced in recent years, new digital platforms driven by artificial intelligence (AI) and machine learning (ML) are increasingly finding practical applications in industry. "Data are so readily available now. Several years ago, we didn't have the manipulation capability, the broad platform or cloud capacity to really work with large volumes of data. We've got that now, so that has been huge in making AI more practical," says Paige Morse, industry marketing director for chemicals at Aspen Technology, Inc. (Bedford, Mass.; www.aspentech.com). While AI and ML have been part of the digitalization discussion for many years, these technologies have not seen a great deal of practical application in the chemical process industries (CPI) until relatively recently, says Don Mack, global alliance manager at Siemens Industry, Inc. (Alpharetta, Ga.; www.industry.usa.siemens.com). "In order for AI to work correctly, it needs data. Control systems and historians in chemical plants have a lot of data available, but in many cases, those data have just been sitting dormant, not really being put to good use. However, new digitalization tools enable us to address some use cases for AI that until recently just weren't possible." This convergence of technologies, from smart sensors to high-performance computing and cloud storage, along with advances in data science, deep learning and access to free and open-source software, have enabled the field of industrial AI to move beyond pure research to practical applications with business benefits, says Samvith Rao, chemical and petroleum industry manager at MathWorks (Natick, Mass.; www.mathworks.com).